Automatic Fairness Testing of Neural Classifiers Through Adversarial Sampling

نویسندگان

چکیده

Although deep learning has demonstrated astonishing performance in many applications, there are still concerns about its dependability. One desirable property of applications with societal impact is fairness (i.e., non-discrimination). Unfortunately, discrimination might be intrinsically embedded into the models due to training data. As a countermeasure, testing systemically identifies discriminatory samples, which can used retrain model and improve model’s fairness. Existing approaches however have two major limitations. First, they only work well on traditional machine poor (e.g., effectiveness efficiency) models. Second, simple structured tabular) data not applicable for domains such as text. In this work, we bridge gap by proposing scalable effective approach systematically searching samples while extending existing address more challenging domain, i.e., text classification. Compared state-of-the-art methods, our employs lightweight procedures like gradient computation clustering, significantly effective. Experimental results show that average, explores search space much effectively (9.62 2.38 times than methods respectively tabular datasets) generates (24.95 2.68 times) within same reasonable time. Moreover, retrained reduce 57.2 60.2 percent average.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Analysing the Fairness of Fairness-aware Classifiers

Calders and Verwer’s two-naive-Bayes is one of fairness-aware classifiers, which classify objects while excluding the influence of a specific information. We analyze why this classifier achieves very high level of the fairness, and show that this is due to a decision rules and a model bias. Based on these findings, we develop methods that are grounded on rigid theory and are applicable to wider...

متن کامل

Automatic Colorization of Grayscale Images Using Generative Adversarial Networks

Automatic colorization of gray scale images poses a unique challenge in Information Retrieval. The goal of this field is to colorize images which have lost some color channels (such as the RGB channels or the AB channels in the LAB color space) while only having the brightness channel available, which is usually the case in a vast array of old photos and portraits. Having the ability to coloriz...

متن کامل

Automatic Testing through Planning

We describe a strategy to automatically test software built using preand postconditions. The strategy searches for valid routine calls: calls for which the preconditions are satisfied. If such calls fail (because the postcondition or another check is violated), we have found a bug. The testing strategy automatically builds a model of the software under test. The model is an abstract version of ...

متن کامل

Software-Defined Adversarial Trajectory Sampling

Today’s routing protocols critically rely on the assumption that the underlying hardware is trusted. Given the increasing number of attacks on network devices, and recent reports on hardware backdoors this assumption has become questionable. Indeed, with the critical role computer networks play today, the contrast between our security assumptions and reality is problematic. This paper presents ...

متن کامل

Sparse Sampling for Adversarial Games

This paper introduces Monte Carlo *-Minimax Search (MCMS), a Monte-Carlo search algorithm for finite, turned based, stochastic, two-player, zero-sum games of perfect information. Through a combination of sparse sampling and classical pruning techniques, MCMS allows deep plans to be constructed. Unlike other popular tree search techniques, MCMS is suitable for densely stochastic games, i.e., gam...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Transactions on Software Engineering

سال: 2022

ISSN: ['0098-5589', '1939-3520', '2326-3881']

DOI: https://doi.org/10.1109/tse.2021.3101478